# Multi-task fine-tuning
Tooka SBERT V2 Small
Tooka-SBERT-V2-Small is a trained sentence transformer model for semantic text similarity and embedding tasks. It can map sentences and paragraphs to a dense vector space where semantically similar texts are close to each other.
Text Embedding
T
PartAI
110
1
Sanskrit Qwen 7B Translate
Apache-2.0
A Sanskrit-specific model fine-tuned based on Qwen2.5-7B, optimized for Sanskrit comprehension and translation
Large Language Model
Transformers

S
diabolic6045
229
2
Qwen2.5 0.5B Portuguese V1
MIT
A Portuguese large language model fine-tuned from Qwen2.5-0.5B-Instruct, specializing in text generation tasks
Large Language Model
Safetensors Other
Q
cnmoro
2,218
4
Modernbert Large Nli
Apache-2.0
A natural language inference model optimized through multi-task fine-tuning based on the ModernBERT-large model, excelling in zero-shot classification and NLI tasks.
Large Language Model
Transformers Supports Multiple Languages

M
tasksource
61.52k
5
Modernbert Base Nli
Apache-2.0
ModernBERT is a model fine-tuned on multi-task natural language inference (NLI) tasks, excelling in zero-shot classification and long-context reasoning.
Large Language Model
Transformers Supports Multiple Languages

M
tasksource
1,867
20
Moxin 7B LLM
Apache-2.0
Moxin 7B is a powerful open-source large language model that offers various types such as base models and chat models, and has demonstrated good performance on multiple common datasets.
Large Language Model
Transformers

M
moxin-org
134
17
Greekbart
MIT
GreekBART is a Greek sequence-to-sequence pre-trained model based on BART, particularly suitable for generation tasks such as summarization.
Large Language Model
Transformers Other

G
dascim
34
0
USER Bge M3
Apache-2.0
Russian universal sentence encoder, based on the sentence-transformers framework, specifically designed to extract 1024-dimensional dense vectors for Russian text
Text Embedding Other
U
deepvk
339.46k
58
Deberta Base Long Nli
Apache-2.0
Based on the DeBERTa-v3-base model, the context length is extended to 1280, and fine-tuned for 250,000 steps on the tasksource dataset, focusing on natural language inference and zero-shot classification tasks.
Large Language Model
Transformers

D
tasksource
541
23
Bert Medium Amharic
A pre-trained Amharic language model based on the bert-medium architecture, with 40.5 million parameters trained on 290 million tokens, achieving performance comparable to larger multilingual models.
Large Language Model
Transformers Other

B
rasyosef
2,661
1
Yi 1.5 34B Chat 16K
Apache-2.0
Yi-1.5 is an upgraded version of the Yi model, demonstrating superior performance in programming, mathematics, reasoning, and instruction-following capabilities.
Large Language Model
Transformers

Y
01-ai
807
27
Yi 1.5 9B
Apache-2.0
Yi-1.5 is an upgraded version of the Yi model, excelling in programming, mathematics, reasoning, and instruction-following capabilities while maintaining excellent language understanding, commonsense reasoning, and reading comprehension.
Large Language Model
Transformers

Y
01-ai
6,140
48
Yi 1.5 9B Chat
Apache-2.0
Yi-1.5 is an upgraded version of the Yi model, excelling in programming, mathematics, reasoning, and instruction-following capabilities while maintaining outstanding language understanding, commonsense reasoning, and reading comprehension.
Large Language Model
Transformers

Y
01-ai
17.16k
143
Akallama Llama3 70b V0.1 GGUF
Other
AkaLlama is a Korean large language model fine-tuned from Meta-Llama-3-70b-Instruct, focusing on multi-task practical applications
Large Language Model Supports Multiple Languages
A
mirlab
414
15
Openelm 3B Instruct
OpenELM is a set of open-source and efficient language models. It adopts a hierarchical parameter allocation strategy to improve model accuracy and includes pre-trained and instruction-tuned versions with 270 million to 3 billion parameters.
Large Language Model
Transformers

O
apple
8,716
333
Configurablesolar 10.7B
Apache-2.0
A configurable large language model fine-tuned using Configurable Safety Tuning (CST) method, supporting behavior pattern configuration through system prompts.
Large Language Model
Transformers

C
vicgalle
1,772
3
Mixtral 8x7B V0.1 Turkish GGUF
Apache-2.0
A model fine-tuned on a specific Turkish dataset, capable of accurately answering information in Turkish and providing strong support for Turkish-related text generation tasks.
Large Language Model
Transformers Supports Multiple Languages

M
sayhan
180
3
Kanarya 750m
Apache-2.0
Canary-750M is a pre-trained Turkish GPT-J 750M model, part of the Turkish Data Depository initiative.
Large Language Model Other
K
asafaya
2,749
10
Vikhrt5 3b
Apache-2.0
Russian-optimized model based on FLAN T5 3b, outperforming FRED T5XL
Large Language Model
Transformers Other

V
Vikhrmodels
35
8
Sentence Camembert Large
Apache-2.0
French sentence embedding model based on CamemBERT-large, providing powerful semantic search capabilities
Text Embedding French
S
Lajavaness
3,729
8
Distilroberta Nli
Apache-2.0
This model is a lightweight natural language inference model based on DistilRoBERTa, supporting zero-shot classification tasks.
Text Classification
Transformers English

D
matekadlicsko
18
0
Deberta V3 Large Zeroshot V1
MIT
A DeBERTa-v3 model specifically designed for zero-shot classification tasks, excelling in various classification tasks
Text Classification
Transformers English

D
MoritzLaurer
10.72k
19
Gpt1
MIT
A Transformer-based language model released by OpenAI, pre-trained on large-scale corpora with powerful text generation capabilities
Large Language Model
Transformers English

G
lgaalves
310
5
Mamba Gpt 3b V4
Apache-2.0
Mamba - GPT - 3B - V4 is an outstanding 3B parameter language model that performs excellently on the Open LLM leaderboard, surpassing dolly - v2 - 12b and providing high - quality language processing capabilities.
Large Language Model
Transformers English

M
CobraMamba
634
8
Camel Platypus2 70B
Camel-Platypus2-70B is a large language model merged from Platypus2-70B and qCammel-70-x, based on the LLaMA 2 architecture, focusing on STEM and logical reasoning tasks.
Large Language Model
Transformers English

C
garage-bAInd
114
15
Tiroberta Abusiveness Detection
A Tigrinya abusive content detection model fine-tuned on TiRoBERTa, trained on 13,717 YouTube comments dataset
Text Classification
Transformers

T
fgaim
210
2
Banglat5 Small
A Bengali pre-trained model based on the sequence-to-sequence Transformer architecture, optimized for natural language generation tasks
Large Language Model
Transformers Other

B
csebuetnlp
510
2
Bert Large Portuguese Cased Legal Mlm Nli Sts V1
MIT
A Portuguese BERT model specialized for the legal domain based on the BERTimbau large model, supporting sentence similarity calculation and semantic search
Text Embedding
Transformers Other

B
stjiris
331
5
Bert Large Portuguese Cased Legal Tsdae Gpl Nli Sts V1
MIT
A legal domain-specific Portuguese sentence transformer based on the BERTimbau large model, supporting semantic similarity calculation
Text Embedding
Transformers Other

B
stjiris
17
0
XLMR MaCoCu Is
XLMR-MaCoCu-is is a large-scale pre-trained language model based on Icelandic text, built by further training the XLM-RoBERTa-large model and belongs to the MaCoCu project.
Large Language Model Other
X
MaCoCu
27
0
XLMR MaltBERTa
A language model based on large-scale pre-training of Maltese text, further trained on the XLM-RoBERTa-large foundation
Large Language Model Other
X
MaCoCu
20
0
Maltberta
MaltBERTa is a large-scale pretrained language model based on Maltese text, using the RoBERTa architecture, developed by the MaCoCu project.
Large Language Model Other
M
MaCoCu
26
0
Banglat5
BanglaT5 is a Bengali sequence-to-sequence transformer model pre-trained with Span Corruption objectives, achieving state-of-the-art performance in multiple Bengali natural language generation tasks.
Large Language Model
Transformers Other

B
csebuetnlp
1,102
15
Kominilm
KoMiniLM is a lightweight Korean language model designed to address latency and capacity limitations of large language models in practical applications.
Large Language Model
Transformers

K
BM-K
244
2
Pko T5 Base
pko-t5 is a T5 model specifically optimized for Korean, trained exclusively on Korean data using BBPE tokenization to address Korean segmentation issues.
Large Language Model
Transformers Korean

P
paust
874
19
Lvbert
Apache-2.0
Latvian pre-trained language model based on BERT architecture, suitable for various natural language understanding tasks
Large Language Model
Transformers Other

L
AiLab-IMCS-UL
473
4
Robbert V2 Dutch Base
MIT
RobBERT is the current state-of-the-art Dutch BERT model, optimized based on the RoBERTa architecture, suitable for various text classification and tagging tasks
Large Language Model Other
R
pdelobelle
7,891
29
It5 Large
Apache-2.0
IT5 is the first family of sequence-to-sequence Transformer models specifically pretrained at scale for Italian, following the T5 model approach.
Large Language Model Other
I
gsarti
37
1
T5 V1 1 Small
Apache-2.0
T5 Version 1.1 is Google's improved text-to-text conversion model, using the GEGLU activation function, pretrained unsupervised only on the C4 dataset, and requires fine-tuning for use.
Large Language Model English
T
google
127.68k
26
Czert B Base Cased
CZERT is a language representation model specifically trained for Czech, outperforming multilingual BERT models on various Czech NLP tasks
Large Language Model
Transformers Other

C
UWB-AIR
560
3
- 1
- 2
Featured Recommended AI Models